Inductive Principles for Learning Restricted Boltzmann Machines (DRAFT: August 25, 2010)
نویسنده
چکیده
We explore the training and usage of the Restricted Boltzmann Machine for unsupervised feature extraction. We investigate the many different aspects involved in their training, and by applying the concept of iterate averaging we show that it is possible to greatly improve on state of the art algorithms. We also derive estimators based on the principles of pseudo-likelihood, ratio matching, and score matching, and we test them empirically against contrastive divergence, and stochastic maximum likelihood (also known as persistent contrastive divergence). Our results show that ratio matching and score matching are promising approaches to learning Restricted Boltzmann Machines. By applying score matching to the Restricted Boltzmann Machine, we show that training an auto-encoder neural network with a particular kind of regularization function is asymptotically consistent. Finally, we discuss the concept of deep learning and its relationship to training Restricted Boltzmann Machines, and briefly explore the impact of fine-tuning on the parameters and performance of a deep belief network.
منابع مشابه
Inductive Principles for Restricted Boltzmann Machine Learning
Recent research has seen the proposal of several new inductive principles designed specifically to avoid the problems associated with maximum likelihood learning in models with intractable partition functions. In this paper, we study learning methods for binary restricted Boltzmann machines (RBMs) based on ratio matching and generalized score matching. We compare these new RBM learning methods ...
متن کاملBiasing Restricted Boltzmann Machines to Manipulate Latent Selectivity and Sparsity
This paper proposes a modification to the restricted Boltzmann machine (RBM) learning algorithm to incorporate inductive biases. These latent activation biases are ideal solutions of the latent activity and may be designed either by modeling neural phenomenon or inductive principles of the task. In this paper, we design activation biases for sparseness and selectivity based on the activation di...
متن کاملAdvances in Deep Learning
Deep neural networks have become increasingly more popular under the name of deep learning recently due to their success in challenging machine learning tasks. Although the popularity is mainly due to the recent successes, the history of neural networks goes as far back as 1958 when Rosenblatt presented a perceptron learning algorithm. Since then, various kinds of artificial neural networks hav...
متن کاملExploiting local structure in stacked Boltzmann machines
Restricted Boltzmann Machines (RBM) are well-studied generative models. For image data, however, standard RBMs are suboptimal, since they do not exploit the local nature of image statistics. We modify RBMs to focus on local structure by restricting visible-hidden interactions. We model long-range interactions using direct or indirect lateral interaction between hidden variables. While learning ...
متن کاملFoundations and Advances in Deep Learning
Aalto University, P.O. Box 11000, FI-00076 Aalto www.aalto.fi Author Kyunghyun Cho Name of the doctoral dissertation Foundations and Advances in Deep Learning Publisher Unit Department of Information and Computer Science Series Aalto University publication series DOCTORAL DISSERTATIONS 21/2014 Field of research Machine Learning Manuscript submitted 2 September 2013 Date of the defence 21 March ...
متن کامل